Equivalence Proofs for Multi-Layer Perceptron Classifiers and the Bayesian Discriminant Function
نویسندگان
چکیده
This paper presents a number of proofs that equate the outputs of a Multi-Layer Perceptron (MLP) classifier and the optimal Bayesian discriminant function for asymptotically large sets of statistically independent training samples. Two broad classes of objective functions are shown to yield Bayesian discriminant performance. The first class are “reasonable error measures,” which achieve Bayesian discriminant performance by engendering classifier outputs that asymptotically equate to a posteriori probabilities. This class includes the mean-squared error (MSE) objective function as well as a number of information theoretic objective functions. The second class are classification figures of merit (CFMmono ), which yield a qualified approximation to Bayesian discriminant performance by engendering classifier outputs that asymptotically identify the maximum a posteriori probability for a given input. Conditions and relationships for Bayesian discriminant functional equivalence are given for both classes of objective functions. Differences between the two classes are then discussed very briefly in the context of how they might affect MLP classifier generalization, given relatively small training sets.
منابع مشابه
Regularization by Early Stopping in Single Layer Perceptron Training
Adaptative training of the non-linear single-layer perceptron can lead to the Euclidean distance classifier and later to the standard Fisher linear discriminant function. On the way between these two classifiers one has a regularized discriminant analysis. That is equivalent to the “weight decay” regularization term added to the cost function. Thus early stopping plays a role of regularization ...
متن کاملA TS Fuzzy Model Derived from a Typical Multi-Layer Perceptron
In this paper, we introduce a Takagi-Sugeno (TS) fuzzy model which is derived from a typical Multi-Layer Perceptron Neural Network (MLP NN). At first, it is shown that the considered MLP NN can be interpreted as a variety of TS fuzzy model. It is discussed that the utilized Membership Function (MF) in such TS fuzzy model, despite its flexible structure, has some major restrictions. After modify...
متن کاملEvolution and generalization of a single neurone: I. Single-layer perceptron as seven statistical classifiers
Unlike many other investigations on this topic, the present one considers the non-linear single-layer perceptron (SLP) as a process in which the weights of the perceptron are increasing, and the cost function of the sum of squares is changing gradually. During the backpropagation training, the decision boundary of of SLP becomes identical or close to that of seven statistical classifiers: (1) t...
متن کاملA Brief Review of Classifiers used in OCR Applications
The performance of a recognition system depends upon the classifiers used for classification purpose. Powerful is the discrimination ability of a classifier, better is its recognition performance. The generalization ability of a classifier is measured on the basis of its performance in classifying the test patterns. There are various factors which affect generalization. Moreover, the feature ex...
متن کاملPinpointing the classifiers of English language writing ability: A discriminant function analysis approach
The major aim of this paper was to investigate the validity of language and intelligence factors for classifying Iranian English learners` writing performance. Iranian participants of the study took three tests for grammar, breadth, and depth of vocabulary, and two tests for verbal and narrative intelligence. They also produced a corpus of argumentative writ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1990